Learning Visual Routines with Reinforcement Learning
نویسنده
چکیده
Reinforcement learning is an ideal framework to learn visual routines since the routines are made up of sequences of actions. However, such algorithms must be able to handle the hidden state (perceptual aliasing) that results from visual routine’s purposefully narrowed attention. The U-Tree algorithm successfully learns visual routines for a complex driving task in which the agent makes eye movements and executes deictic actions in order to weave in and out of traffic on a four-laned highway. The task involves hidden state, time pressure, stochasticity, a large world state space, and a large perceptual state space. U-Tree uses a tree-structured representation, and is related to work on Prediction Suffix Trees (Ron, Singer, &; Tishby 1994), Parti-game (Moore 1993), algorithm (Chapman 8z Kaelbling 1991), and Variable Resolution Dynamic Programming (Moore 1991). Tree is a direct descendant of Utile Suffix Memory (McCallum 1995c), which used short-term memory, but not selective perception. Unlike Whitehead’s Lion algorithm, the algorithm handles noise, large state spaces, and uses short-term memory to uncover hidden state.
منابع مشابه
Reinforcement Learning of a Simple Visual Task
Raw visual data must be interpreted in some fashion to result in spatial perception. Some models for this interpretation hypothesize visual routines as underlying mechanisms for distilling the properties of a scene. Visual routines must determine these properties quickly without higher–level information. We present an approach for learning an example of a visual routine using reinforcement. Rat...
متن کاملLearning to Look in Different Environments: An Active-Vision Model Which Learns and Readapts Visual Routines
One of the main claims of the active vision framework is that finding data on the basis of task requirements is more efficient than reconstructing the whole scene by performing a complete visual scan. To be successful, this approach requires that agents learn visual routines to direct overt attention to locations with the information needed to accomplish the task. In ecological conditions, lear...
متن کاملEfficient Exploration in Reinforcement Learning with Hidden State
Undoubtedly, efficient exploration is crucial for the success of a learning agent. Previous approaches to exploration in reinforcement learning exclusively address exploration in Markovian domains, i.e. domains in which the state of the environment is fully observable. If the environment is only partially observable, they cease to work because exploration statistics are confounded between alias...
متن کاملMotion Sketch: Acquisition of Visual Motion Guided Behaviors
Sensor and motor systems are not separable for autonomous agents to accomplish tasks in a dynamic environment. This paper proposes a method to represent the interaction between a vision-based learning agent and its environment. The method is called \motion sketch" by which a one-eyed mobile robot can learn several behaviors such as obstacle avoidance and target pursuit. A motion sketch is a col...
متن کاملActive Gesture Recognition using Learned Visual Attention
We have developed a foveated gesture recognition system that runs in an unconstrained office environment with an active camera. Using vision routines previously implemented for an interactive environment, we determine the spatial location of salient body parts of a user and guide an active camera to obtain images of gestures or expressions. A hidden-state reinforcement learning paradigm is used...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 1996